Complexity Preserving Functions
نویسنده
چکیده
It’s widely recognized that compression is useful and even necessary for inductive learning, where a short description will capture the ‘regularities’. We introduce complexitypreserving functions that preserve these regularities of the concept. They are based on the universal information distance [Bennett et al. ‘97] and define for an instance a set of elements sharing the same complexity type. This corresponds to the two-part code [Rissanen ‘89] of the MDL principle, when it is interpreted as the first term describing the set and the second term the element in the set [Rissanen 99]. We investigate its importance in inductive learning. Introduction The Kolmogorov complexity of a string x K(x) is the length of a shortest program to compute x on a universal computer. It represents the minimal amount of information required to generate x by an effective process [Kolmogorov 65]. Bennett et al. studied the definition of a universal cognitive distance [Bennett et al. ‘97], based on the observation that “the psychological similarity between two objects x and y is the complexity of the simplest transformation between them” [see also Chater 2003]. They start by defining the shortest program p to compute y from x: KF= min{l(p) : F(p, x) = y} Where F(p,x) is a partial recursive function, that can be computed on a Turing machine and l(p) denotes the (binary) length of vector p. To attain a symmetric definition, Benett et al. investigate reversible computations, where the overlap between the information to convert x into y and y into x is maximized. We try to define functions, by using the notion of information distance, that not only preserve complexity, but also capture the ‘complexity type’ or ‘regularity’ of an object. K-preserving functions A Kolmogorov complexity-preserving function gx is a one-to-one reversible function defined from (X, R) -> X: x’ = gx(x, k). Except for a limited number of special k values, where the K-complexity is reduced in x’, what we call the reduction values, the following constraints apply for all k: 1. x = gx(x’, -k). 2. K(gx,k|x) < K(gx, k), x contains information to perform gx. 3. The program p to compute gx is both a minimal program, for computing x’ from x and vice versa. This last condition means that there is no extra information needed for either computation, there is maximum overlap of information and the complexity is similar in x and x’. Hence, the difference between the complexities of x and x’, K(x) and K(x’), depends only on k, not on gx. The differences between the complexities is bounded by |K(x) K(x’)| ≤ l(k). There is a fourth constraint on the implementation of the k-values: 4. gx (x, a.k + b.l) = a.gx(b.gx(x, l), k) Examples of K-preserving functions are translations, rotations, enlargements, repetitions, polynomials, permutations,... On a circle, translations in each direction and resizing preserve the complexity. An ellipse can additionally be stretched along both its axes, however, for a special reduction value k, the ellipse becomes a circle, that is of lower complexity. Set with the same regularities By applying all existing gx functions on x by varying the k values, we become a set Cx of elements with the same regularity. Each element of this set got the same K-preserving functions and will generate the same set (following from constraint 4). This means that they all share the same regularity. There are dependencies among the Kpreserving functions. If for a certain x: gx,1(x, k1) = gx,3(gx,2(x, k2), k3), (with all k nonreduction values), gx,1 can be composed out of gx,2 and gx,3 for all elements of Cx. After elimination of the dependent gx,i, a set of independent K-preserving functions Cg{gx,i} remains, which represents the degrees of freedom of the regularity. The set Cx got Πki elements and each element in the set is represented by its (k1, k2, ...), after having chosen a reference element (with all k’s set to
منابع مشابه
Linear Functions Preserving Sut-Majorization on RN
Suppose $textbf{M}_{n}$ is the vector space of all $n$-by-$n$ real matrices, and let $mathbb{R}^{n}$ be the set of all $n$-by-$1$ real vectors. A matrix $Rin textbf{M}_{n}$ is said to be $textit{row substochastic}$ if it has nonnegative entries and each row sum is at most $1$. For $x$, $y in mathbb{R}^{n}$, it is said that $x$ is $textit{sut-majorized}$ by $y$ (denoted by $ xprec_{sut} y$) if t...
متن کاملTENSION QUARTIC TRIGONOMETRIC BÉZIER CURVES PRESERVING INTERPOLATION CURVES SHAPE
In this paper simple quartic trigonometric polynomial blending functions, with a tensionparameter, are presented. These type of functions are useful for constructing trigonometricB´ezier curves and surfaces, they can be applied to construct continuous shape preservinginterpolation spline curves with shape parameters. To better visualize objects and graphics atension parameter is included. In th...
متن کاملShift Invariant Spaces and Shift Preserving Operators on Locally Compact Abelian Groups
We investigate shift invariant subspaces of $L^2(G)$, where $G$ is a locally compact abelian group. We show that every shift invariant space can be decomposed as an orthogonal sum of spaces each of which is generated by a single function whose shifts form a Parseval frame. For a second countable locally compact abelian group $G$ we prove a useful Hilbert space isomorphism, introduce range funct...
متن کاملLinear Functions Preserving Multivariate and Directional Majorization
Let V and W be two real vector spaces and let &sim be a relation on both V and W. A linear function T : V → W is said to be a linear preserver (respectively strong linear preserver) of &sim if Tx &sim Ty whenever x &sim y (respectively Tx &sim Ty if and only if x &sim y). In this paper we characterize all linear functions T : M_{n,m} → M_{n,k} which preserve or strongly preserve multivariate an...
متن کاملState of the Art in Similarity Preserving Hashing Functions
One of the goals of digital forensics is to analyse the content of digital devices by reducing its size and complexity. Similarity preserving hashing functions help to accomplish that mission through a resemblance comparison between different files. Some of the best-known functions of this type are the context-triggered piecewise hashing functions, which create a signature formed by several has...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2003